candidate entity
- North America > United States (0.45)
- Europe > United Kingdom (0.28)
- North America > Canada (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Transportation > Passenger (1.00)
- Transportation > Infrastructure & Services (1.00)
- Transportation > Ground > Road (1.00)
- (16 more...)
Leveraging the Power of Large Language Models in Entity Linking via Adaptive Routing and Targeted Reasoning
Li, Yajie, Galimov, Albert, Ganapaneni, Mitra Datta, Thejaswi, Pujitha, Meng, De, Kumar, Priyanshu, Potdar, Saloni
Entity Linking (EL) has traditionally relied on large annotated datasets and extensive model fine-tuning. While recent few-shot methods leverage large language models (LLMs) through prompting to reduce training requirements, they often suffer from inefficiencies due to expensive LLM-based reasoning. ARTER (Adaptive Routing and Targeted Entity Reasoning) presents a structured pipeline that achieves high performance without deep fine-tuning by strategically combining candidate generation, context-based scoring, adaptive routing, and selective reasoning. ARTER computes a small set of complementary signals(both embedding and LLM-based) over the retrieved candidates to categorize contextual mentions into easy and hard cases. The cases are then handled by a low-computational entity linker (e.g. ReFinED) and more expensive targeted LLM-based reasoning respectively. On standard benchmarks, ARTER outperforms ReFinED by up to +4.47%, with an average gain of +2.53% on 5 out of 6 datasets, and performs comparably to pipelines using LLM-based reasoning for all mentions, while being as twice as efficient in terms of the number of LLM tokens.
- Oceania > Australia (0.04)
- North America > Canada (0.04)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.04)
- (9 more...)
Harnessing Deep LLM Participation for Robust Entity Linking
Hou, Jiajun, Zhang, Chenyu, Meng, Rui
Entity Linking (EL), the task of mapping textual entity mentions to their corresponding entries in knowledge bases, constitutes a fundamental component of natural language understanding. Recent advancements in Large Language Models (LLMs) have demonstrated remarkable potential for enhancing EL performance. Prior research has leveraged LLMs to improve entity disambiguation and input representation, yielding significant gains in accuracy and robustness. However, these approaches typically apply LLMs to isolated stages of the EL task, failing to fully integrate their capabilities throughout the entire process. In this work, we introduce DeepEL, a comprehensive framework that incorporates LLMs into every stage of the entity linking task. Furthermore, we identify that disambiguating entities in isolation is insufficient for optimal performance. To address this limitation, we propose a novel self-validation mechanism that utilizes global contextual information, enabling LLMs to rectify their own predictions and better recognize cohesive relationships among entities within the same sentence. Extensive empirical evaluation across ten benchmark datasets demonstrates that DeepEL substantially outperforms existing state-of-the-art methods, achieving an average improvement of 2.6\% in overall F1 score and a remarkable 4% gain on out-of-domain datasets. These results underscore the efficacy of deep LLM integration in advancing the state-of-the-art in entity linking.
- Europe > France (0.05)
- North America > United States > District of Columbia > Washington (0.04)
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.04)
- (5 more...)
- Research Report > New Finding (0.66)
- Research Report > Promising Solution (0.48)
DBLPLink 2.0 -- An Entity Linker for the DBLP Scholarly Knowledge Graph
Banerjee, Debayan, Taffa, Tilahun Abedissa, Usbeck, Ricardo
In this work we present an entity linker for DBLP's 2025 version of RDF-based Knowledge Graph. Compared to the 2022 version, DBLP now considers publication venues as a new entity type called dblp:Stream. In the earlier version of DBLPLink, we trained KG-embeddings and re-rankers on a dataset to produce entity linkings. In contrast, in this work, we develop a zero-shot entity linker using LLMs using a novel method, where we re-rank candidate entities based on the log-probabilities of the "yes" token output at the penultimate layer of the LLM.
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Spain > Catalonia > Barcelona Province > Barcelona (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- (3 more...)
Generative Annotation for ASR Named Entity Correction
Luo, Yuanchang, Wei, Daimeng, Li, Shaojun, Shang, Hengchao, Guo, Jiaxin, Li, Zongyao, Wu, Zhanglin, Chen, Xiaoyu, Rao, Zhiqiang, Yang, Jinlong, Yang, Hao
End-to-end automatic speech recognition systems often fail to transcribe domain-specific named entities, causing catastrophic failures in downstream tasks. Numerous fast and lightweight named entity correction (NEC) models have been proposed in recent years. These models, mainly leveraging phonetic-level edit distance algorithms, have shown impressive performances. However, when the forms of the wrongly-transcribed words(s) and the ground-truth entity are significantly different, these methods often fail to locate the wrongly transcribed words in hypothesis, thus limiting their usage. We propose a novel NEC method that utilizes speech sound features to retrieve candidate entities. With speech sound features and candidate entities, we inovatively design a generative method to annotate entity errors in ASR transcripts and replace the text with correct entities. This method is effective in scenarios of word form difference. We test our method using open-source and self-constructed test sets. The results demonstrate that our NEC method can bring significant improvement to entity accuracy. The self-constructed training data and test set is publicly available at github.com/L6-NLP/Generative-Annotation-NEC.
A Unified Biomedical Named Entity Recognition Framework with Large Language Models
Lv, Tengxiao, Luo, Ling, Li, Juntao, Wang, Yanhua, Pan, Yuchen, Liu, Chao, Wang, Yanan, Jiang, Yan, Lv, Huiyi, Sun, Yuanyuan, Wang, Jian, Lin, Hongfei
Accurate recognition of biomedical named entities is critical for medical information extraction and knowledge discovery. However, existing methods often struggle with nested entities, entity boundary ambiguity, and cross-lingual generalization. In this paper, we propose a unified Biomedical Named Entity Recognition (BioNER) framework based on Large Language Models (LLMs). We first reformulate BioNER as a text generation task and design a symbolic tagging strategy to jointly handle both flat and nested entities with explicit boundary annotation. To enhance multilingual and multi-task generalization, we perform bilingual joint fine-tuning across multiple Chinese and English datasets. Additionally, we introduce a contrastive learning-based entity selector that filters incorrect or spurious predictions by leveraging boundary-sensitive positive and negative samples. Experimental results on four benchmark datasets and two unseen corpora show that our method achieves state-of-the-art performance and robust zero-shot generalization across languages. The source codes are freely available at https://github.com/dreamer-tx/LLMNER.
- Europe > United Kingdom (0.28)
- North America > United States > Texas > Loving County (0.14)
- South America > Brazil > Federal District > Brasília (0.04)
- (6 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Transportation > Passenger (1.00)
- Transportation > Infrastructure & Services (1.00)
- Transportation > Ground > Road (1.00)
- (16 more...)
PGMEL: Policy Gradient-based Generative Adversarial Network for Multimodal Entity Linking
Pooja, KM, Long, Cheng, Sun, Aixin
Abstract--The task of entity linking, which involves associating mentions with their respective entities in a knowledge graph, has received significant attention due to its numerous potential applications. Recently, various multimodal entity linking (MEL) techniques have been proposed, targeted to learn comprehensive embeddings by leveraging both text and vision modalities. The selection of high-quality negative samples can potentially play a crucial role in metric/representation learning. However, to the best of our knowledge, this possibility remains unexplored in existing literature within the framework of MEL. T o fill this gap, we address the multimodal entity linking problem in a generative adversarial setting where the generator is responsible for generating high-quality negative samples, and the discriminator is assigned the responsibility for the metric learning tasks. Since the generator is involved in generating samples, which is a discrete process, we optimize it using policy gradient techniques and propose a policy gradient-based generative adversarial network for multimodal entity linking (PGMEL). Experimental results based on Wiki-MEL, Richpedia-MEL and WikiDiverse datasets demonstrate that PGMEL learns meaningful representation by selecting challenging negative samples and outperforms state-of-the-art methods. The last few decades have seen unprecedented growth in data availability. However, the increasing data availability quickly becomes a liability rather than an asset due to the increased gap between data and information. Thus, information extraction (IE) techniques to retrieve knowledge/information from a large amount of data have received considerable attention recently. A knowledge graph (KG) is a structured information database that allows storing extracted information from a large amount of data for retrieval or reasoning at a later stage. Furthermore, the recent developments in IE techniques allow the automatic creation of large KGs with millions of entries from knowledge bases such as Wikipedia, DBpedia, Freebase, and Y AGO [1]. Automated KG construction is a complex task that involves various intricate subtasks, including (i) named entity recognition to identify and categorize named entities, like a person or geographic locations, etc., in text, (ii) co-reference resolution to group references of the same entity, (iii) relation extraction to establish relationships between the entities, and (iv) entity linking [2], [3]. KM Pooja is with the Department of Information Technology, Indian Institute of Information Technology, Allahabad India 211012. Cheng Long and Aixin Sun are with the School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, Singapore 639798.
Guided Navigation in Knowledge-Dense Environments: Structured Semantic Exploration with Guidance Graphs
Tao, Dehao, Liu, Guangjie, Weizheng, null, Huang, Yongfeng, jiang, Minghu
While Large Language Models (LLMs) exhibit strong linguistic capabilities, their reliance on static knowledge and opaque reasoning processes limits their performance in knowledge-intensive tasks. Knowledge graphs (KGs) offer a promising solution, but current exploration methods face a fundamental trade-off: question-guided approaches incur redundant exploration due to granularity mismatches, while clue-guided methods fail to effectively leverage contextual information for complex scenarios. To address these limitations, we propose Guidance-Graph-guided Knowledge Exploration (GG-Explore), a novel framework that introduces an intermediate Guidance Graph to bridge unstructured queries and structured knowledge retrieval. The Guidance Graph defines the retrieval space by abstracting the target knowledge's structure while preserving broader semantic context, enabling precise and efficient exploration. Building upon the Guidance Graph, we develop: (1) Structural Alignment that filters incompatible candidates without LLM overhead, and (2) Context-A ware Pruning that enforces semantic consistency with graph constraints. Extensive experiments show our method achieves superior efficiency and outperforms SOT A, especially on complex tasks, while maintaining strong performance with smaller LLMs, demonstrating practical value.
An Entity Linking Agent for Question Answering
Luo, Yajie, Wu, Yihong, Li, Muzhi, Mo, Fengran, Sun, Jia Ao, Wang, Xinyu, Ma, Liheng, Zhang, Yingxue, Nie, Jian-Yun
Some Question Answering (QA) systems rely on knowledge bases (KBs) to provide accurate answers. Entity Linking (EL) plays a critical role in linking natural language mentions to KB entries. However, most existing EL methods are designed for long contexts and do not perform well on short, ambiguous user questions in QA tasks. We propose an entity linking agent for QA, based on a Large Language Model that simulates human cognitive workflows. The agent actively identifies entity mentions, retrieves candidate entities, and makes decision. To verify the effectiveness of our agent, we conduct two experiments: tool-based entity linking and QA task evaluation. The results confirm the robustness and effectiveness of our agent.
- North America > United States > New Mexico > Bernalillo County > Albuquerque (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > China > Hong Kong (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.96)
- (2 more...)